An analytic comparison of regularization methods for Gaussian Processes
نویسندگان
چکیده
Gaussian Processes (GPs) are often used to predict the output of a parameterized deterministic experiment. They have many applications in the field of Computer Experiments, in particular to perform sensitivity analysis, adaptive design of experiments and global optimization. Nearly all of the applications of GPs to Computer Experiments require the inversion of a covariance matrix. Because this matrix is often ill-conditioned, regularization techniques are required. Today, there is still a need to better regularize GPs. The two most classical regularization methods are i) pseudoinverse (PI) and ii) nugget (or jitter or observation noise). This article provides algebraic calculations which allow comparing PI and nugget regularizations. It is proven that pseudoinverse regularization averages the output values and makes the variance null at redundant points. On the opposite, nugget regularization lacks interpolation properties but preserves a non-zero variance at every point. However, these two regularization techniques become similar as the nugget value decreases. A distribution-wise GP is introduced which interpolates Gaussian distributions instead of data points and mitigates the drawbacks of pseudoinverse and nugget regularized GPs. Finally, data-model discrepancy is discussed and serves as a guide for choosing a regularization technique.
منابع مشابه
A dynamical system perspective of structural learning with forgetting
Structural learning with forgetting is an established method of using Laplace regularization to generate skeletal artificial neural networks. In this paper we develop a continuous dynamical system model of regularization in which the associated regularization parameter is generalized to be a time-varying function. Analytic results are obtained for a Laplace regularizer and a quadratic error sur...
متن کاملRegularization with Dot-Product Kernels
In this paper we give necessary and sufficient conditions under which kernels of dot product type k(x, y) = k(x · y) satisfy Mercer’s condition and thus may be used in Support Vector Machines (SVM), Regularization Networks (RN) or Gaussian Processes (GP). In particular, we show that if the kernel is analytic (i.e. can be expanded in a Taylor series), all expansion coefficients have to be nonneg...
متن کاملComprehensive Multi-Criteria Comparison and Ranking of Natural Gas Liquefaction Process by Analytic Hierarchy Process (AHP)
Several processes have been proposed for natural gas liquefaction due to the vast utilization of LNG as a reliable and relatively easy to use fuel. Even though the merits and demerits of different process have been studied, a dearth of comprehensive technical and economical comparative investigation of these methods makes further broad examination a necessity. This article is presented to addre...
متن کاملNorm Regularization Algorithm for Image Deconvolution
Up to now, the non-convex l p (0 < p < 1) norm regularization function has shown good performance for sparse signal processing. Indeed, it benefits from a significantly heavier-tailed hyper-Laplacian model, which is desirable in the context of image gradient distributions. Both l 1/2 and l 2/3 regularization methods have been given analytic solutions and fast closed-form thresholding formulae i...
متن کاملThe Rate of Entropy for Gaussian Processes
In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...
متن کامل